variational algorithm - traducción al ruso
Diclib.com
Diccionario ChatGPT
Ingrese una palabra o frase en cualquier idioma 👆
Idioma:

Traducción y análisis de palabras por inteligencia artificial ChatGPT

En esta página puede obtener un análisis detallado de una palabra o frase, producido utilizando la mejor tecnología de inteligencia artificial hasta la fecha:

  • cómo se usa la palabra
  • frecuencia de uso
  • se utiliza con más frecuencia en el habla oral o escrita
  • opciones de traducción
  • ejemplos de uso (varias frases con traducción)
  • etimología

variational algorithm - traducción al ruso

MATHEMATICAL METHODS USED IN BAYESIAN INFERENCE AND MACHINE LEARNING
Variational bayes; Variational Bayes; Variational Bayesian method; Variational inference; Variational free energy
  • Pictorial illustration of coordinate ascent variational inference algorithm by the duality formula<ref name=Yoon2021/>

variational algorithm      

математика

вариационный алгоритм

minimum principle         
DIFFERENTIAL CALCULUS ON FUNCTION SPACES
Variational methods; Variational calculus; Variational Calculus; Minimum principle; Calculus of variation; Strong extrema; Variational Method; Calculus of Variations; Applications of the calculus of variations; Variational method; Variation calculus; Variational problem; Variational formulation

математика

принцип минимума

algorithm         
  • Alan Turing's statue at [[Bletchley Park]]
  • The example-diagram of Euclid's algorithm from T.L. Heath (1908), with more detail added. Euclid does not go beyond a third measuring and gives no numerical examples. Nicomachus gives the example of 49 and 21: "I subtract the less from the greater; 28 is left; then again I subtract from this the same 21 (for this is possible); 7 is left; I subtract this from 21, 14 is left; from which I again subtract 7 (for this is possible); 7 is left, but 7 cannot be subtracted from 7." Heath comments that "The last phrase is curious, but the meaning of it is obvious enough, as also the meaning of the phrase about ending 'at one and the same number'."(Heath 1908:300).
  • "Inelegant" is a translation of Knuth's version of the algorithm with a subtraction-based remainder-loop replacing his use of division (or a "modulus" instruction). Derived from Knuth 1973:2–4. Depending on the two numbers "Inelegant" may compute the g.c.d. in fewer steps than "Elegant".
  • 1=IF test THEN GOTO step xxx}}, shown as diamond), the unconditional GOTO (rectangle), various assignment operators (rectangle), and HALT (rectangle). Nesting of these structures inside assignment-blocks results in complex diagrams (cf. Tausworthe 1977:100, 114).
  • A graphical expression of Euclid's algorithm to find the greatest common divisor for 1599 and 650
<syntaxhighlight lang="text" highlight="1,5">
 1599 = 650×2 + 299
 650 = 299×2 + 52
 299 = 52×5 + 39
 52 = 39×1 + 13
 39 = 13×3 + 0</syntaxhighlight>
SEQUENCE OF INSTRUCTIONS TO PERFORM A TASK
Algorithmically; Computer algorithm; Properties of algorithms; Algorithim; Algoritmi de Numero Indorum; Algoritmi de numero indorum; Algoritmi De Numero Indorum; Алгоритм; Algorithem; Software logic; Computer algorithms; Encoding Algorithm; Naive algorithm; Naïve algorithm; Algorithm design; Algorithm segment; Algorithmic problem; Algorythm; Rule set; Continuous algorithm; Algorithms; Software-based; Algorithmic method; Algorhthym; Algorthym; Algorhythms; Formalization of algorithms; Mathematical algorithm; Draft:GE8151 Problem Solving and Python Programming; Computational algorithms; Optimization algorithms; Algorithm classification; History of algorithms; Patented algorithms; Algorithmus
algorithm noun math. алгоритм algorithm validation - проверка правильности алгоритма

Definición

Euclidean Algorithm

Wikipedia

Variational Bayesian methods

Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables (usually termed "data") as well as unknown parameters and latent variables, with various sorts of relationships among the three types of random variables, as might be described by a graphical model. As typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables". Variational Bayesian methods are primarily used for two purposes:

  1. To provide an analytical approximation to the posterior probability of the unobserved variables, in order to do statistical inference over these variables.
  2. To derive a lower bound for the marginal likelihood (sometimes called the evidence) of the observed data (i.e. the marginal probability of the data given the model, with marginalization performed over unobserved variables). This is typically used for performing model selection, the general idea being that a higher marginal likelihood for a given model indicates a better fit of the data by that model and hence a greater probability that the model in question was the one that generated the data. (See also the Bayes factor article.)

In the former purpose (that of approximating a posterior probability), variational Bayes is an alternative to Monte Carlo sampling methods—particularly, Markov chain Monte Carlo methods such as Gibbs sampling—for taking a fully Bayesian approach to statistical inference over complex distributions that are difficult to evaluate directly or sample. In particular, whereas Monte Carlo techniques provide a numerical approximation to the exact posterior using a set of samples, variational Bayes provides a locally-optimal, exact analytical solution to an approximation of the posterior.

Variational Bayes can be seen as an extension of the expectation-maximization (EM) algorithm from maximum a posteriori estimation (MAP estimation) of the single most probable value of each parameter to fully Bayesian estimation which computes (an approximation to) the entire posterior distribution of the parameters and latent variables. As in EM, it finds a set of optimal parameter values, and it has the same alternating structure as does EM, based on a set of interlocked (mutually dependent) equations that cannot be solved analytically.

For many applications, variational Bayes produces solutions of comparable accuracy to Gibbs sampling at greater speed. However, deriving the set of equations used to update the parameters iteratively often requires a large amount of work compared with deriving the comparable Gibbs sampling equations. This is the case even for many models that are conceptually quite simple, as is demonstrated below in the case of a basic non-hierarchical model with only two parameters and no latent variables.

¿Cómo se dice variational algorithm en Ruso? Traducción de &#39variational algorithm&#39 al Ruso